# Multilingual pre-training
Codegen 350m Html
Bsd-3-clause
CodeGen-HTML 350M is an autoregressive language model fine-tuned from CodeGen-Multi 350M, specifically designed for HTML code generation.
Large Language Model
Transformers Other

C
alecsharpie
132
14
Wav2vec2 Large Xlsr 53 Toy Train Data Augment 0.1.csv
Apache-2.0
This model is a speech recognition model fine-tuned from facebook/wav2vec2-base, trained using data augmentation techniques
Speech Recognition
Transformers

W
scasutt
22
0
Wav2vec2 Xlsr Multilingual 53 Fa
A multilingual speech recognition model based on the wav2vec 2.0 architecture, specifically fine-tuned for Persian, significantly reducing word error rate
Speech Recognition
Transformers

W
masoudmzb
83
7
Phoneme Test 5 Sv
Apache-2.0
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MULTILINGUAL_LIBRISPEECH - German 10-hour dataset for German speech recognition tasks.
Speech Recognition
Transformers

P
patrickvonplaten
17
0
Wav2vec2 Large Xlsr Turkish Demo
This model is an XLSR-Wav2Vec2 speech recognition model fine-tuned on the Turkish Common Voice dataset, primarily used for Turkish speech-to-text tasks.
Speech Recognition
W
patrickvonplaten
18
0
Tf Xlm Roberta Base
XLM-RoBERTa is an extended version of a cross-lingual sentence encoder, trained on 2.5T of data across 100 languages, achieving excellent performance in multiple cross-lingual benchmarks.
Large Language Model
Transformers

T
jplu
4,820
1
Wav2vec2 Base 10k Voxpopuli Ft Fi
An automatic speech recognition model based on Facebook's Wav2Vec2 base model, pre-trained on a 10K unlabeled subset of the VoxPopuli corpus and fine-tuned on Finnish transcription data.
Speech Recognition
Transformers Other

W
facebook
24
0
Wav2vec2 Base 10k Voxpopuli Ft Cs
A speech recognition model based on Facebook's Wav2Vec2 architecture, pre-trained with 10K unlabeled Czech data from the VoxPopuli corpus and fine-tuned on Czech transcription data.
Speech Recognition
Transformers Other

W
facebook
226
0
Wav2vec2 Base 10k Voxpopuli Ft Pl
Pre-trained on 10K unlabeled data from the VoxPopuli corpus and fine-tuned on Polish transcription data
Speech Recognition
Transformers Other

W
facebook
203
3
Cino Base V2
Apache-2.0
CINO is a multilingual pre-trained model designed for Chinese minority languages, supporting Chinese and 7 minority languages, built on the XLM-R framework.
Large Language Model
Transformers Supports Multiple Languages

C
hfl
156
5
Mbart Large 50 Many To Many Mmt
A multilingual machine translation model fine-tuned based on mBART-large-50, supporting translation between 50 languages
Machine Translation Supports Multiple Languages
M
facebook
404.66k
357
Wav2vec2 Base 10k Voxpopuli Ft Sk
Pre-trained on 10K hours of unlabeled VoxPopuli corpus data and fine-tuned on Slovak transcription data
Speech Recognition
Transformers Other

W
facebook
39
1
Featured Recommended AI Models